Zipf1 New Model Test Data Result Combined

Result

Model Summaries

Model Better than base % of the times
LR_10[cache_size=0.001,treshold=0.3] 0
LR_10[cache_size=0.001,treshold=0.5] 0
LR_10[cache_size=0.001,treshold=0.6] 0
LR_10[cache_size=0.001,treshold=0.7] 0
LR_10[cache_size=0.001,treshold=0.8] 0
LR_10[cache_size=0.001,treshold=0.9] 100
LR_10[cache_size=All,treshold=0.3] 0
LR_10[cache_size=All,treshold=0.5] 0
LR_10[cache_size=All,treshold=0.6] 0
LR_10[cache_size=All,treshold=0.7] 0
LR_10[cache_size=All,treshold=0.8] 0
LR_10[cache_size=All,treshold=0.9] 0
LR_11[cache_size=0.001,treshold=0.3] 0
LR_11[cache_size=0.001,treshold=0.5] 0
LR_11[cache_size=0.001,treshold=0.6] 0
LR_11[cache_size=0.001,treshold=0.7] 0
LR_11[cache_size=0.001,treshold=0.8] 0
LR_11[cache_size=0.001,treshold=0.9] 100
LR_11[cache_size=All,treshold=0.3] 0
LR_11[cache_size=All,treshold=0.5] 0
LR_11[cache_size=All,treshold=0.6] 0
LR_11[cache_size=All,treshold=0.7] 0
LR_11[cache_size=All,treshold=0.8] 0
LR_11[cache_size=All,treshold=0.9] 0
LR_12[cache_size=0.001,treshold=0.3] 0
LR_12[cache_size=0.001,treshold=0.5] 0
LR_12[cache_size=0.001,treshold=0.6] 0
LR_12[cache_size=0.001,treshold=0.7] 0
LR_12[cache_size=0.001,treshold=0.8] 0
LR_12[cache_size=0.001,treshold=0.9] 100
LR_12[cache_size=All,treshold=0.3] 0
LR_12[cache_size=All,treshold=0.5] 0
LR_12[cache_size=All,treshold=0.6] 0
LR_12[cache_size=All,treshold=0.7] 0
LR_12[cache_size=All,treshold=0.8] 0
LR_12[cache_size=All,treshold=0.9] 0
LR_13[cache_size=0.001,treshold=0.3] 0
LR_13[cache_size=0.001,treshold=0.5] 0
LR_13[cache_size=0.001,treshold=0.6] 0
LR_13[cache_size=0.001,treshold=0.7] 0
LR_13[cache_size=0.001,treshold=0.8] 0
LR_13[cache_size=0.001,treshold=0.9] 0
LR_13[cache_size=All,treshold=0.3] 0
LR_13[cache_size=All,treshold=0.5] 20
LR_13[cache_size=All,treshold=0.6] 0
LR_13[cache_size=All,treshold=0.7] 0
LR_13[cache_size=All,treshold=0.8] 0
LR_13[cache_size=All,treshold=0.9] 0
LR_14[cache_size=0.001,treshold=0.3] 0
LR_14[cache_size=0.001,treshold=0.5] 0
LR_14[cache_size=0.001,treshold=0.6] 0
LR_14[cache_size=0.001,treshold=0.7] 0
LR_14[cache_size=0.001,treshold=0.8] 0
LR_14[cache_size=0.001,treshold=0.9] 0
LR_14[cache_size=All,treshold=0.3] 0
LR_14[cache_size=All,treshold=0.5] 20
LR_14[cache_size=All,treshold=0.6] 0
LR_14[cache_size=All,treshold=0.7] 0
LR_14[cache_size=All,treshold=0.8] 0
LR_14[cache_size=All,treshold=0.9] 0
LR_15[cache_size=0.001,treshold=0.3] 0
LR_15[cache_size=0.001,treshold=0.5] 0
LR_15[cache_size=0.001,treshold=0.6] 0
LR_15[cache_size=0.001,treshold=0.7] 0
LR_15[cache_size=0.001,treshold=0.8] 0
LR_15[cache_size=0.001,treshold=0.9] 0
LR_15[cache_size=All,treshold=0.3] 0
LR_15[cache_size=All,treshold=0.5] 20
LR_15[cache_size=All,treshold=0.6] 0
LR_15[cache_size=All,treshold=0.7] 0
LR_15[cache_size=All,treshold=0.8] 0
LR_15[cache_size=All,treshold=0.9] 0
LR_7[cache_size=0.001,treshold=0.3] 0
LR_7[cache_size=0.001,treshold=0.5] 0
LR_7[cache_size=0.001,treshold=0.6] 0
LR_7[cache_size=0.001,treshold=0.7] 0
LR_7[cache_size=0.001,treshold=0.8] 0
LR_7[cache_size=0.001,treshold=0.9] 100
LR_7[cache_size=All,treshold=0.3] 0
LR_7[cache_size=All,treshold=0.5] 0
LR_7[cache_size=All,treshold=0.6] 0
LR_7[cache_size=All,treshold=0.7] 0
LR_7[cache_size=All,treshold=0.8] 0
LR_7[cache_size=All,treshold=0.9] 0
LR_8[cache_size=0.001,treshold=0.3] 0
LR_8[cache_size=0.001,treshold=0.5] 0
LR_8[cache_size=0.001,treshold=0.6] 0
LR_8[cache_size=0.001,treshold=0.7] 0
LR_8[cache_size=0.001,treshold=0.8] 0
LR_8[cache_size=0.001,treshold=0.9] 100
LR_8[cache_size=All,treshold=0.3] 0
LR_8[cache_size=All,treshold=0.5] 0
LR_8[cache_size=All,treshold=0.6] 0
LR_8[cache_size=All,treshold=0.7] 0
LR_8[cache_size=All,treshold=0.8] 0
LR_8[cache_size=All,treshold=0.9] 0
LR_9[cache_size=0.001,treshold=0.3] 0
LR_9[cache_size=0.001,treshold=0.5] 0
LR_9[cache_size=0.001,treshold=0.6] 0
LR_9[cache_size=0.001,treshold=0.7] 0
LR_9[cache_size=0.001,treshold=0.8] 0
LR_9[cache_size=0.001,treshold=0.9] 100
LR_9[cache_size=All,treshold=0.3] 0
LR_9[cache_size=All,treshold=0.5] 0
LR_9[cache_size=All,treshold=0.6] 0
LR_9[cache_size=All,treshold=0.7] 0
LR_9[cache_size=All,treshold=0.8] 0
LR_9[cache_size=All,treshold=0.9] 0
LR_10[cache_size=0.01,treshold=0.3] 0
LR_10[cache_size=0.01,treshold=0.5] 0
LR_10[cache_size=0.01,treshold=0.6] 0
LR_10[cache_size=0.01,treshold=0.7] 0
LR_10[cache_size=0.01,treshold=0.8] 0
LR_10[cache_size=0.01,treshold=0.9] 0
LR_11[cache_size=0.01,treshold=0.3] 0
LR_11[cache_size=0.01,treshold=0.5] 0
LR_11[cache_size=0.01,treshold=0.6] 0
LR_11[cache_size=0.01,treshold=0.7] 0
LR_11[cache_size=0.01,treshold=0.8] 0
LR_11[cache_size=0.01,treshold=0.9] 0
LR_12[cache_size=0.01,treshold=0.3] 0
LR_12[cache_size=0.01,treshold=0.5] 0
LR_12[cache_size=0.01,treshold=0.6] 0
LR_12[cache_size=0.01,treshold=0.7] 0
LR_12[cache_size=0.01,treshold=0.8] 0
LR_12[cache_size=0.01,treshold=0.9] 0
LR_13[cache_size=0.01,treshold=0.3] 0
LR_13[cache_size=0.01,treshold=0.5] 0
LR_13[cache_size=0.01,treshold=0.6] 0
LR_13[cache_size=0.01,treshold=0.7] 0
LR_13[cache_size=0.01,treshold=0.8] 0
LR_13[cache_size=0.01,treshold=0.9] 0
LR_14[cache_size=0.01,treshold=0.3] 0
LR_14[cache_size=0.01,treshold=0.5] 0
LR_14[cache_size=0.01,treshold=0.6] 0
LR_14[cache_size=0.01,treshold=0.7] 0
LR_14[cache_size=0.01,treshold=0.8] 0
LR_14[cache_size=0.01,treshold=0.9] 0
LR_15[cache_size=0.01,treshold=0.3] 0
LR_15[cache_size=0.01,treshold=0.5] 0
LR_15[cache_size=0.01,treshold=0.6] 0
LR_15[cache_size=0.01,treshold=0.7] 0
LR_15[cache_size=0.01,treshold=0.8] 0
LR_15[cache_size=0.01,treshold=0.9] 0
LR_7[cache_size=0.01,treshold=0.3] 0
LR_7[cache_size=0.01,treshold=0.5] 0
LR_7[cache_size=0.01,treshold=0.6] 0
LR_7[cache_size=0.01,treshold=0.7] 0
LR_7[cache_size=0.01,treshold=0.8] 0
LR_7[cache_size=0.01,treshold=0.9] 100
LR_8[cache_size=0.01,treshold=0.3] 0
LR_8[cache_size=0.01,treshold=0.5] 0
LR_8[cache_size=0.01,treshold=0.6] 0
LR_8[cache_size=0.01,treshold=0.7] 0
LR_8[cache_size=0.01,treshold=0.8] 0
LR_8[cache_size=0.01,treshold=0.9] 100
LR_9[cache_size=0.01,treshold=0.3] 0
LR_9[cache_size=0.01,treshold=0.5] 0
LR_9[cache_size=0.01,treshold=0.6] 0
LR_9[cache_size=0.01,treshold=0.7] 0
LR_9[cache_size=0.01,treshold=0.8] 0
LR_9[cache_size=0.01,treshold=0.9] 100
LR_10[cache_size=0.1,treshold=0.3] 0
LR_10[cache_size=0.1,treshold=0.5] 0
LR_10[cache_size=0.1,treshold=0.6] 0
LR_10[cache_size=0.1,treshold=0.7] 0
LR_10[cache_size=0.1,treshold=0.8] 0
LR_10[cache_size=0.1,treshold=0.9] 0
LR_11[cache_size=0.1,treshold=0.3] 0
LR_11[cache_size=0.1,treshold=0.5] 0
LR_11[cache_size=0.1,treshold=0.6] 0
LR_11[cache_size=0.1,treshold=0.7] 0
LR_11[cache_size=0.1,treshold=0.8] 0
LR_11[cache_size=0.1,treshold=0.9] 0
LR_12[cache_size=0.1,treshold=0.3] 0
LR_12[cache_size=0.1,treshold=0.5] 0
LR_12[cache_size=0.1,treshold=0.6] 0
LR_12[cache_size=0.1,treshold=0.7] 0
LR_12[cache_size=0.1,treshold=0.8] 0
LR_12[cache_size=0.1,treshold=0.9] 0
LR_13[cache_size=0.1,treshold=0.3] 0
LR_13[cache_size=0.1,treshold=0.5] 0
LR_13[cache_size=0.1,treshold=0.6] 0
LR_13[cache_size=0.1,treshold=0.7] 0
LR_13[cache_size=0.1,treshold=0.8] 0
LR_13[cache_size=0.1,treshold=0.9] 0
LR_14[cache_size=0.1,treshold=0.3] 0
LR_14[cache_size=0.1,treshold=0.5] 0
LR_14[cache_size=0.1,treshold=0.6] 0
LR_14[cache_size=0.1,treshold=0.7] 0
LR_14[cache_size=0.1,treshold=0.8] 0
LR_14[cache_size=0.1,treshold=0.9] 0
LR_15[cache_size=0.1,treshold=0.3] 0
LR_15[cache_size=0.1,treshold=0.5] 0
LR_15[cache_size=0.1,treshold=0.6] 0
LR_15[cache_size=0.1,treshold=0.7] 0
LR_15[cache_size=0.1,treshold=0.8] 0
LR_15[cache_size=0.1,treshold=0.9] 0
LR_7[cache_size=0.1,treshold=0.3] 0
LR_7[cache_size=0.1,treshold=0.5] 0
LR_7[cache_size=0.1,treshold=0.6] 0
LR_7[cache_size=0.1,treshold=0.7] 0
LR_7[cache_size=0.1,treshold=0.8] 0
LR_7[cache_size=0.1,treshold=0.9] 0
LR_8[cache_size=0.1,treshold=0.3] 0
LR_8[cache_size=0.1,treshold=0.5] 0
LR_8[cache_size=0.1,treshold=0.6] 0
LR_8[cache_size=0.1,treshold=0.7] 0
LR_8[cache_size=0.1,treshold=0.8] 0
LR_8[cache_size=0.1,treshold=0.9] 0
LR_9[cache_size=0.1,treshold=0.3] 0
LR_9[cache_size=0.1,treshold=0.5] 0
LR_9[cache_size=0.1,treshold=0.6] 0
LR_9[cache_size=0.1,treshold=0.7] 0
LR_9[cache_size=0.1,treshold=0.8] 0
LR_9[cache_size=0.1,treshold=0.9] 0
LR_10[cache_size=0.2,treshold=0.3] 0
LR_10[cache_size=0.2,treshold=0.5] 0
LR_10[cache_size=0.2,treshold=0.6] 0
LR_10[cache_size=0.2,treshold=0.7] 0
LR_10[cache_size=0.2,treshold=0.8] 0
LR_10[cache_size=0.2,treshold=0.9] 0
LR_11[cache_size=0.2,treshold=0.3] 0
LR_11[cache_size=0.2,treshold=0.5] 0
LR_11[cache_size=0.2,treshold=0.6] 0
LR_11[cache_size=0.2,treshold=0.7] 0
LR_11[cache_size=0.2,treshold=0.8] 0
LR_11[cache_size=0.2,treshold=0.9] 0
LR_12[cache_size=0.2,treshold=0.3] 0
LR_12[cache_size=0.2,treshold=0.5] 0
LR_12[cache_size=0.2,treshold=0.6] 0
LR_12[cache_size=0.2,treshold=0.7] 0
LR_12[cache_size=0.2,treshold=0.8] 0
LR_12[cache_size=0.2,treshold=0.9] 0
LR_13[cache_size=0.2,treshold=0.3] 0
LR_13[cache_size=0.2,treshold=0.5] 0
LR_13[cache_size=0.2,treshold=0.6] 0
LR_13[cache_size=0.2,treshold=0.7] 0
LR_13[cache_size=0.2,treshold=0.8] 0
LR_13[cache_size=0.2,treshold=0.9] 0
LR_14[cache_size=0.2,treshold=0.3] 0
LR_14[cache_size=0.2,treshold=0.5] 0
LR_14[cache_size=0.2,treshold=0.6] 0
LR_14[cache_size=0.2,treshold=0.7] 0
LR_14[cache_size=0.2,treshold=0.8] 0
LR_14[cache_size=0.2,treshold=0.9] 0
LR_15[cache_size=0.2,treshold=0.3] 0
LR_15[cache_size=0.2,treshold=0.5] 0
LR_15[cache_size=0.2,treshold=0.6] 0
LR_15[cache_size=0.2,treshold=0.7] 0
LR_15[cache_size=0.2,treshold=0.8] 0
LR_15[cache_size=0.2,treshold=0.9] 0
LR_7[cache_size=0.2,treshold=0.3] 0
LR_7[cache_size=0.2,treshold=0.5] 0
LR_7[cache_size=0.2,treshold=0.6] 0
LR_7[cache_size=0.2,treshold=0.7] 0
LR_7[cache_size=0.2,treshold=0.8] 0
LR_7[cache_size=0.2,treshold=0.9] 0
LR_8[cache_size=0.2,treshold=0.3] 0
LR_8[cache_size=0.2,treshold=0.5] 0
LR_8[cache_size=0.2,treshold=0.6] 0
LR_8[cache_size=0.2,treshold=0.7] 0
LR_8[cache_size=0.2,treshold=0.8] 0
LR_8[cache_size=0.2,treshold=0.9] 0
LR_9[cache_size=0.2,treshold=0.3] 0
LR_9[cache_size=0.2,treshold=0.5] 0
LR_9[cache_size=0.2,treshold=0.6] 0
LR_9[cache_size=0.2,treshold=0.7] 0
LR_9[cache_size=0.2,treshold=0.8] 0
LR_9[cache_size=0.2,treshold=0.9] 0
LR_10[cache_size=0.4,treshold=0.3] 0
LR_10[cache_size=0.4,treshold=0.5] 0
LR_10[cache_size=0.4,treshold=0.6] 0
LR_10[cache_size=0.4,treshold=0.7] 0
LR_10[cache_size=0.4,treshold=0.8] 0
LR_10[cache_size=0.4,treshold=0.9] 0
LR_11[cache_size=0.4,treshold=0.3] 0
LR_11[cache_size=0.4,treshold=0.5] 0
LR_11[cache_size=0.4,treshold=0.6] 0
LR_11[cache_size=0.4,treshold=0.7] 0
LR_11[cache_size=0.4,treshold=0.8] 0
LR_11[cache_size=0.4,treshold=0.9] 0
LR_12[cache_size=0.4,treshold=0.3] 0
LR_12[cache_size=0.4,treshold=0.5] 0
LR_12[cache_size=0.4,treshold=0.6] 0
LR_12[cache_size=0.4,treshold=0.7] 0
LR_12[cache_size=0.4,treshold=0.8] 0
LR_12[cache_size=0.4,treshold=0.9] 0
LR_13[cache_size=0.4,treshold=0.3] 0
LR_13[cache_size=0.4,treshold=0.5] 0
LR_13[cache_size=0.4,treshold=0.6] 0
LR_13[cache_size=0.4,treshold=0.7] 0
LR_13[cache_size=0.4,treshold=0.8] 0
LR_13[cache_size=0.4,treshold=0.9] 0
LR_14[cache_size=0.4,treshold=0.3] 0
LR_14[cache_size=0.4,treshold=0.5] 0
LR_14[cache_size=0.4,treshold=0.6] 0
LR_14[cache_size=0.4,treshold=0.7] 0
LR_14[cache_size=0.4,treshold=0.8] 0
LR_14[cache_size=0.4,treshold=0.9] 0
LR_15[cache_size=0.4,treshold=0.3] 0
LR_15[cache_size=0.4,treshold=0.5] 0
LR_15[cache_size=0.4,treshold=0.6] 0
LR_15[cache_size=0.4,treshold=0.7] 0
LR_15[cache_size=0.4,treshold=0.8] 0
LR_15[cache_size=0.4,treshold=0.9] 0
LR_7[cache_size=0.4,treshold=0.3] 0
LR_7[cache_size=0.4,treshold=0.5] 0
LR_7[cache_size=0.4,treshold=0.6] 0
LR_7[cache_size=0.4,treshold=0.7] 0
LR_7[cache_size=0.4,treshold=0.8] 0
LR_7[cache_size=0.4,treshold=0.9] 0
LR_8[cache_size=0.4,treshold=0.3] 0
LR_8[cache_size=0.4,treshold=0.5] 0
LR_8[cache_size=0.4,treshold=0.6] 0
LR_8[cache_size=0.4,treshold=0.7] 0
LR_8[cache_size=0.4,treshold=0.8] 0
LR_8[cache_size=0.4,treshold=0.9] 0
LR_9[cache_size=0.4,treshold=0.3] 0
LR_9[cache_size=0.4,treshold=0.5] 0
LR_9[cache_size=0.4,treshold=0.6] 0
LR_9[cache_size=0.4,treshold=0.7] 0
LR_9[cache_size=0.4,treshold=0.8] 0
LR_9[cache_size=0.4,treshold=0.9] 0
Offline Clock 1st iteration 0
Offline Clock 2nd iteration 100
Zipf Optimal Distribution 100

Promotion Reduced (%)

Model Max Min Avg Mdn
LR_10[cache_size=0.001,treshold=0.3] 64.487 64.4568 64.4714 64.4716
LR_10[cache_size=0.001,treshold=0.5] 53.0073 52.9553 52.9846 52.9966
LR_10[cache_size=0.001,treshold=0.6] 45.4336 45.3647 45.3964 45.3967
LR_10[cache_size=0.001,treshold=0.7] 35.2239 35.1603 35.1838 35.1825
LR_10[cache_size=0.001,treshold=0.8] 20.19 20.1546 20.1658 20.1625
LR_10[cache_size=0.001,treshold=0.9] 0.0489568 0.0458181 0.0475212 0.0474412
LR_10[cache_size=All,treshold=0.3] 99.9133 65.2775 91.592 99.2999
LR_10[cache_size=All,treshold=0.5] 99.6467 8.24613 75.4498 97.1811
LR_10[cache_size=All,treshold=0.6] 0 0 0 0
LR_10[cache_size=All,treshold=0.7] 0 0 0 0
LR_10[cache_size=All,treshold=0.8] 0 0 0 0
LR_10[cache_size=All,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.001,treshold=0.3] 64.9619 64.9316 64.9501 64.9501
LR_11[cache_size=0.001,treshold=0.5] 53.307 53.2581 53.2857 53.3009
LR_11[cache_size=0.001,treshold=0.6] 45.4136 45.3526 45.3804 45.3735
LR_11[cache_size=0.001,treshold=0.7] 34.6595 34.6152 34.6336 34.6296
LR_11[cache_size=0.001,treshold=0.8] 19.2793 19.241 19.2604 19.2604
LR_11[cache_size=0.001,treshold=0.9] 0.811724 0.806042 0.809589 0.809746
LR_11[cache_size=All,treshold=0.3] 99.9133 65.2759 91.5911 99.2995
LR_11[cache_size=All,treshold=0.5] 99.6478 8.28082 75.4773 97.1887
LR_11[cache_size=All,treshold=0.6] 0 0 0 0
LR_11[cache_size=All,treshold=0.7] 0 0 0 0
LR_11[cache_size=All,treshold=0.8] 0 0 0 0
LR_11[cache_size=All,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.001,treshold=0.3] 64.6 64.5723 64.586 64.5864
LR_12[cache_size=0.001,treshold=0.5] 52.9772 52.9285 52.9559 52.9715
LR_12[cache_size=0.001,treshold=0.6] 45.3007 45.2448 45.2688 45.2616
LR_12[cache_size=0.001,treshold=0.7] 35.0088 34.9565 34.9749 34.9686
LR_12[cache_size=0.001,treshold=0.8] 20.1304 20.0888 20.1081 20.1093
LR_12[cache_size=0.001,treshold=0.9] 0.614198 0.608547 0.611838 0.611802
LR_12[cache_size=All,treshold=0.3] 99.9136 65.269 91.5943 99.3023
LR_12[cache_size=All,treshold=0.5] 99.6462 8.19283 75.4266 97.1788
LR_12[cache_size=All,treshold=0.6] 0 0 0 0
LR_12[cache_size=All,treshold=0.7] 0 0 0 0
LR_12[cache_size=All,treshold=0.8] 0 0 0 0
LR_12[cache_size=All,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.001,treshold=0.3] 66.829 66.8063 66.818 66.8192
LR_13[cache_size=0.001,treshold=0.5] 54.9514 54.9081 54.9317 54.9384
LR_13[cache_size=0.001,treshold=0.6] 46.7527 46.7059 46.727 46.7234
LR_13[cache_size=0.001,treshold=0.7] 35.2222 35.1685 35.1912 35.1846
LR_13[cache_size=0.001,treshold=0.8] 17.4747 17.4552 17.4597 17.4559
LR_13[cache_size=0.001,treshold=0.9] 0 0 0 0
LR_13[cache_size=All,treshold=0.3] 99.9777 67.5232 92.8336 99.8203
LR_13[cache_size=All,treshold=0.5] 99.7419 0.606482 74.5379 98.1443
LR_13[cache_size=All,treshold=0.6] 0 0 0 0
LR_13[cache_size=All,treshold=0.7] 0 0 0 0
LR_13[cache_size=All,treshold=0.8] 0 0 0 0
LR_13[cache_size=All,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.001,treshold=0.3] 67.3954 67.371 67.3851 67.3879
LR_14[cache_size=0.001,treshold=0.5] 55.1991 55.1603 55.1758 55.1762
LR_14[cache_size=0.001,treshold=0.6] 46.0919 46.0451 46.0657 46.055
LR_14[cache_size=0.001,treshold=0.7] 33.9776 33.9478 33.9622 33.9624
LR_14[cache_size=0.001,treshold=0.8] 16.4262 16.3975 16.4127 16.4109
LR_14[cache_size=0.001,treshold=0.9] 0.00016319 5.76087e-05 0.000113279 0.000115193
LR_14[cache_size=All,treshold=0.3] 99.9772 67.7593 92.8719 99.8181
LR_14[cache_size=All,treshold=0.5] 99.7279 0.561001 74.0434 98.0193
LR_14[cache_size=All,treshold=0.6] 0 0 0 0
LR_14[cache_size=All,treshold=0.7] 0 0 0 0
LR_14[cache_size=All,treshold=0.8] 0 0 0 0
LR_14[cache_size=All,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.001,treshold=0.3] 66.963 66.9422 66.9539 66.9532
LR_15[cache_size=0.001,treshold=0.5] 54.8713 54.8295 54.8497 54.8541
LR_15[cache_size=0.001,treshold=0.6] 46.5126 46.461 46.4868 46.4836
LR_15[cache_size=0.001,treshold=0.7] 34.9055 34.8619 34.88 34.8766
LR_15[cache_size=0.001,treshold=0.8] 17.5543 17.5221 17.5365 17.5342
LR_15[cache_size=0.001,treshold=0.9] 0 0 0 0
LR_15[cache_size=All,treshold=0.3] 99.9774 67.7008 92.8671 99.8203
LR_15[cache_size=All,treshold=0.5] 99.7205 0.509145 74.0405 97.9911
LR_15[cache_size=All,treshold=0.6] 0 0 0 0
LR_15[cache_size=All,treshold=0.7] 0 0 0 0
LR_15[cache_size=All,treshold=0.8] 0 0 0 0
LR_15[cache_size=All,treshold=0.9] 0 0 0 0
LR_7[cache_size=0.001,treshold=0.3] 69.3839 69.3614 69.3742 69.3748
LR_7[cache_size=0.001,treshold=0.5] 55.0665 55.0078 55.0368 55.0426
LR_7[cache_size=0.001,treshold=0.6] 44.7721 44.7306 44.7498 44.7505
LR_7[cache_size=0.001,treshold=0.7] 30.4216 30.3858 30.4051 30.4041
LR_7[cache_size=0.001,treshold=0.8] 10.7908 10.7669 10.7769 10.7716
LR_7[cache_size=0.001,treshold=0.9] 2.08557 2.06788 2.08039 2.08282
LR_7[cache_size=All,treshold=0.3] 99.2371 66.7883 92.212 99.1317
LR_7[cache_size=All,treshold=0.5] 96.6911 6.58047 74.107 95.2891
LR_7[cache_size=All,treshold=0.6] 0 0 0 0
LR_7[cache_size=All,treshold=0.7] 0 0 0 0
LR_7[cache_size=All,treshold=0.8] 0 0 0 0
LR_7[cache_size=All,treshold=0.9] 0 0 0 0
LR_8[cache_size=0.001,treshold=0.3] 70.2804 70.264 70.2712 70.2708
LR_8[cache_size=0.001,treshold=0.5] 54.6318 54.584 54.6119 54.6171
LR_8[cache_size=0.001,treshold=0.6] 43.4294 43.3924 43.414 43.4207
LR_8[cache_size=0.001,treshold=0.7] 29.4672 29.433 29.454 29.4546
LR_8[cache_size=0.001,treshold=0.8] 11.7258 11.7036 11.7137 11.7074
LR_8[cache_size=0.001,treshold=0.9] 1.95965 1.93887 1.95192 1.95418
LR_8[cache_size=All,treshold=0.3] 99.2281 67.5488 92.3776 99.1269
LR_8[cache_size=All,treshold=0.5] 96.5693 5.51532 72.8871 94.775
LR_8[cache_size=All,treshold=0.6] 0 0 0 0
LR_8[cache_size=All,treshold=0.7] 0 0 0 0
LR_8[cache_size=All,treshold=0.8] 0 0 0 0
LR_8[cache_size=All,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.001,treshold=0.3] 69.4067 69.3914 69.4012 69.4034
LR_9[cache_size=0.001,treshold=0.5] 54.7783 54.719 54.7512 54.7566
LR_9[cache_size=0.001,treshold=0.6] 44.4575 44.4136 44.435 44.4393
LR_9[cache_size=0.001,treshold=0.7] 30.3545 30.3241 30.3427 30.3407
LR_9[cache_size=0.001,treshold=0.8] 11.4069 11.3961 11.401 11.399
LR_9[cache_size=0.001,treshold=0.9] 2.0927 2.07646 2.08734 2.08817
LR_9[cache_size=All,treshold=0.3] 99.2329 67.4621 92.3623 99.1293
LR_9[cache_size=All,treshold=0.5] 96.602 4.77935 72.7267 94.7841
LR_9[cache_size=All,treshold=0.6] 0 0 0 0
LR_9[cache_size=All,treshold=0.7] 0 0 0 0
LR_9[cache_size=All,treshold=0.8] 0 0 0 0
LR_9[cache_size=All,treshold=0.9] 0 0 0 0
LR_10[cache_size=0.01,treshold=0.3] 67.8627 67.8449 67.8553 67.8553
LR_10[cache_size=0.01,treshold=0.5] 55.4986 55.4549 55.4816 55.4801
LR_10[cache_size=0.01,treshold=0.6] 46.7602 46.7276 46.7433 46.7427
LR_10[cache_size=0.01,treshold=0.7] 34.4885 34.4587 34.4708 34.4681
LR_10[cache_size=0.01,treshold=0.8] 16.5256 16.4739 16.5061 16.5117
LR_10[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.01,treshold=0.3] 67.8952 67.8824 67.8897 67.8909
LR_11[cache_size=0.01,treshold=0.5] 55.3983 55.3555 55.3813 55.3809
LR_11[cache_size=0.01,treshold=0.6] 46.5881 46.5561 46.5707 46.57
LR_11[cache_size=0.01,treshold=0.7] 34.3995 34.3685 34.3824 34.3775
LR_11[cache_size=0.01,treshold=0.8] 16.5874 16.5366 16.5689 16.5762
LR_11[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.01,treshold=0.3] 67.838 67.8195 67.8298 67.8316
LR_12[cache_size=0.01,treshold=0.5] 55.5696 55.5221 55.549 55.5451
LR_12[cache_size=0.01,treshold=0.6] 46.8503 46.8153 46.8322 46.8322
LR_12[cache_size=0.01,treshold=0.7] 34.5398 34.5128 34.5247 34.5231
LR_12[cache_size=0.01,treshold=0.8] 16.4081 16.3551 16.386 16.3901
LR_12[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.01,treshold=0.3] 74.0248 74.0001 74.0165 74.0177
LR_13[cache_size=0.01,treshold=0.5] 58.5534 58.5279 58.5398 58.5403
LR_13[cache_size=0.01,treshold=0.6] 46.9892 46.9501 46.9768 46.9833
LR_13[cache_size=0.01,treshold=0.7] 32.4144 32.3608 32.3965 32.4004
LR_13[cache_size=0.01,treshold=0.8] 7.82267 7.80438 7.81444 7.81283
LR_13[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.01,treshold=0.3] 73.989 73.964 73.9815 73.9852
LR_14[cache_size=0.01,treshold=0.5] 58.2852 58.2595 58.2743 58.2714
LR_14[cache_size=0.01,treshold=0.6] 46.806 46.7635 46.7913 46.7958
LR_14[cache_size=0.01,treshold=0.7] 32.3267 32.2724 32.307 32.3086
LR_14[cache_size=0.01,treshold=0.8] 8.32671 8.30491 8.31782 8.31693
LR_14[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.01,treshold=0.3] 74.0456 74.0226 74.0384 74.0432
LR_15[cache_size=0.01,treshold=0.5] 58.6472 58.6192 58.6329 58.633
LR_15[cache_size=0.01,treshold=0.6] 47.0552 47.0175 47.043 47.0512
LR_15[cache_size=0.01,treshold=0.7] 32.4456 32.3889 32.4258 32.4263
LR_15[cache_size=0.01,treshold=0.8] 7.64665 7.62795 7.63708 7.63426
LR_15[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_7[cache_size=0.01,treshold=0.3] 68.8959 68.8611 68.8799 68.8839
LR_7[cache_size=0.01,treshold=0.5] 54.3901 54.3474 54.3671 54.3647
LR_7[cache_size=0.01,treshold=0.6] 44.4902 44.4516 44.4727 44.4755
LR_7[cache_size=0.01,treshold=0.7] 31.1963 31.1736 31.1881 31.1898
LR_7[cache_size=0.01,treshold=0.8] 13.1008 13.0903 13.095 13.0926
LR_7[cache_size=0.01,treshold=0.9] 1.55324 1.54283 1.54949 1.55177
LR_8[cache_size=0.01,treshold=0.3] 68.8875 68.8517 68.8704 68.872
LR_8[cache_size=0.01,treshold=0.5] 54.3587 54.3158 54.3351 54.3318
LR_8[cache_size=0.01,treshold=0.6] 44.4638 44.4248 44.4453 44.4475
LR_8[cache_size=0.01,treshold=0.7] 31.1961 31.1752 31.1887 31.189
LR_8[cache_size=0.01,treshold=0.8] 13.146 13.1354 13.141 13.1401
LR_8[cache_size=0.01,treshold=0.9] 1.54754 1.53787 1.54443 1.54725
LR_9[cache_size=0.01,treshold=0.3] 69.0317 68.9917 69.0127 69.0161
LR_9[cache_size=0.01,treshold=0.5] 54.6412 54.5996 54.6155 54.6095
LR_9[cache_size=0.01,treshold=0.6] 44.6607 44.6221 44.6421 44.6479
LR_9[cache_size=0.01,treshold=0.7] 31.1167 31.0942 31.1081 31.1102
LR_9[cache_size=0.01,treshold=0.8] 12.6351 12.6235 12.6298 12.6296
LR_9[cache_size=0.01,treshold=0.9] 1.54677 1.5375 1.54328 1.54577
LR_10[cache_size=0.1,treshold=0.3] 74.9456 74.9101 74.9288 74.9259
LR_10[cache_size=0.1,treshold=0.5] 62.6098 62.5507 62.584 62.5891
LR_10[cache_size=0.1,treshold=0.6] 51.9497 51.9058 51.9295 51.9351
LR_10[cache_size=0.1,treshold=0.7] 21.6587 21.6172 21.6426 21.6488
LR_10[cache_size=0.1,treshold=0.8] 0.0109785 0.00986522 0.0103329 0.0100063
LR_10[cache_size=0.1,treshold=0.9] 0.00753458 0.00630925 0.00684968 0.00671965
LR_11[cache_size=0.1,treshold=0.3] 74.9828 74.9523 74.9681 74.9644
LR_11[cache_size=0.1,treshold=0.5] 63.0481 62.9884 63.0237 63.0299
LR_11[cache_size=0.1,treshold=0.6] 52.305 52.261 52.2842 52.284
LR_11[cache_size=0.1,treshold=0.7] 20.7968 20.7638 20.7788 20.7795
LR_11[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.1,treshold=0.3] 74.9838 74.9532 74.969 74.9648
LR_12[cache_size=0.1,treshold=0.5] 63.3339 63.2755 63.3056 63.3126
LR_12[cache_size=0.1,treshold=0.6] 52.51 52.4658 52.4853 52.4828
LR_12[cache_size=0.1,treshold=0.7] 20.444 20.4068 20.4248 20.4304
LR_12[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_12[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.1,treshold=0.3] 75.0436 75.0123 75.025 75.0142
LR_13[cache_size=0.1,treshold=0.5] 56.5272 56.4923 56.5077 56.5129
LR_13[cache_size=0.1,treshold=0.6] 43.2788 43.2088 43.2477 43.2431
LR_13[cache_size=0.1,treshold=0.7] 23.8681 23.839 23.8544 23.8565
LR_13[cache_size=0.1,treshold=0.8] 0.00270622 0.00188888 0.00226051 0.00223988
LR_13[cache_size=0.1,treshold=0.9] 0.000214161 0.000136341 0.00018886 0.000194675
LR_14[cache_size=0.1,treshold=0.3] 75.2085 75.1766 75.1906 75.1814
LR_14[cache_size=0.1,treshold=0.5] 56.5957 56.564 56.579 56.5836
LR_14[cache_size=0.1,treshold=0.6] 43.2875 43.2164 43.2557 43.2506
LR_14[cache_size=0.1,treshold=0.7] 23.6399 23.6103 23.6243 23.6245
LR_14[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_14[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.1,treshold=0.3] 75.2561 75.2234 75.2383 75.2306
LR_15[cache_size=0.1,treshold=0.5] 56.6606 56.6298 56.6441 56.6489
LR_15[cache_size=0.1,treshold=0.6] 43.2663 43.1944 43.2335 43.23
LR_15[cache_size=0.1,treshold=0.7] 23.5367 23.5091 23.5222 23.524
LR_15[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_15[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_7[cache_size=0.1,treshold=0.3] 68.923 68.8922 68.9102 68.9191
LR_7[cache_size=0.1,treshold=0.5] 54.548 54.4748 54.5193 54.5296
LR_7[cache_size=0.1,treshold=0.6] 44.6964 44.6353 44.6727 44.6912
LR_7[cache_size=0.1,treshold=0.7] 31.3573 31.2821 31.311 31.2999
LR_7[cache_size=0.1,treshold=0.8] 12.5234 12.4943 12.5057 12.5024
LR_7[cache_size=0.1,treshold=0.9] 0.00770981 0.00649424 0.00700548 0.00685185
LR_8[cache_size=0.1,treshold=0.3] 69.0532 69.0238 69.041 69.0467
LR_8[cache_size=0.1,treshold=0.5] 54.6622 54.5917 54.6359 54.6467
LR_8[cache_size=0.1,treshold=0.6] 44.7504 44.6907 44.7268 44.7461
LR_8[cache_size=0.1,treshold=0.7] 31.2825 31.2146 31.2384 31.2272
LR_8[cache_size=0.1,treshold=0.8] 12.2738 12.2486 12.2584 12.258
LR_8[cache_size=0.1,treshold=0.9] 0.00202563 0.001441 0.00170174 0.00156697
LR_9[cache_size=0.1,treshold=0.3] 69.0755 69.0468 69.0634 69.0689
LR_9[cache_size=0.1,treshold=0.5] 54.6932 54.6235 54.6677 54.6782
LR_9[cache_size=0.1,treshold=0.6] 44.7754 44.7194 44.7534 44.7722
LR_9[cache_size=0.1,treshold=0.7] 31.2972 31.2263 31.2513 31.24
LR_9[cache_size=0.1,treshold=0.8] 12.2528 12.2285 12.2387 12.2386
LR_9[cache_size=0.1,treshold=0.9] 0.00206459 0.00149942 0.00175042 0.00162537
LR_10[cache_size=0.2,treshold=0.3] 74.4583 74.4314 74.4462 74.449
LR_10[cache_size=0.2,treshold=0.5] 56.1385 56.1264 56.1318 56.1288
LR_10[cache_size=0.2,treshold=0.6] 42.9998 42.9608 42.9763 42.9755
LR_10[cache_size=0.2,treshold=0.7] 23.5909 23.5487 23.5701 23.5765
LR_10[cache_size=0.2,treshold=0.8] 0.00441873 0.00392863 0.00420447 0.00421954
LR_10[cache_size=0.2,treshold=0.9] 0.00297914 0.00256911 0.00273501 0.00270989
LR_11[cache_size=0.2,treshold=0.3] 74.5336 74.51 74.5207 74.5214
LR_11[cache_size=0.2,treshold=0.5] 56.1767 56.1642 56.1701 56.1677
LR_11[cache_size=0.2,treshold=0.6] 42.9901 42.9512 42.9672 42.9654
LR_11[cache_size=0.2,treshold=0.7] 23.4647 23.4244 23.4461 23.4549
LR_11[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.2,treshold=0.3] 74.5689 74.5431 74.5558 74.5573
LR_12[cache_size=0.2,treshold=0.5] 56.193 56.1809 56.1865 56.1838
LR_12[cache_size=0.2,treshold=0.6] 42.9788 42.9402 42.9555 42.9518
LR_12[cache_size=0.2,treshold=0.7] 23.4095 23.3668 23.3908 23.4003
LR_12[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_12[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.2,treshold=0.3] 74.5525 74.5278 74.5402 74.5423
LR_13[cache_size=0.2,treshold=0.5] 56.1839 56.1722 56.1774 56.174
LR_13[cache_size=0.2,treshold=0.6] 42.9849 42.9466 42.9625 42.9601
LR_13[cache_size=0.2,treshold=0.7] 23.4421 23.4012 23.4231 23.4328
LR_13[cache_size=0.2,treshold=0.8] 7.9977e-05 9.99959e-06 3.19886e-05 1.99978e-05
LR_13[cache_size=0.2,treshold=0.9] 9.99653e-06 0 1.99931e-06 0
LR_14[cache_size=0.2,treshold=0.3] 74.5338 74.5102 74.521 74.5217
LR_14[cache_size=0.2,treshold=0.5] 56.1768 56.1643 56.1703 56.1678
LR_14[cache_size=0.2,treshold=0.6] 42.9902 42.9513 42.9673 42.9654
LR_14[cache_size=0.2,treshold=0.7] 23.4647 23.4244 23.4461 23.455
LR_14[cache_size=0.2,treshold=0.8] 9.99959e-06 0 5.99901e-06 9.99653e-06
LR_14[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.2,treshold=0.3] 74.5683 74.5423 74.555 74.5567
LR_15[cache_size=0.2,treshold=0.5] 56.1948 56.1823 56.1884 56.1859
LR_15[cache_size=0.2,treshold=0.6] 42.9813 42.9424 42.9582 42.9546
LR_15[cache_size=0.2,treshold=0.7] 23.4121 23.3698 23.3935 23.4032
LR_15[cache_size=0.2,treshold=0.8] 0.000109968 9.98949e-06 6.59849e-05 6.99757e-05
LR_15[cache_size=0.2,treshold=0.9] 3.99885e-05 0 1.99951e-05 1.99978e-05
LR_7[cache_size=0.2,treshold=0.3] 69.5891 69.5646 69.5797 69.5878
LR_7[cache_size=0.2,treshold=0.5] 54.8336 54.7859 54.8112 54.8093
LR_7[cache_size=0.2,treshold=0.6] 44.5147 44.4682 44.4938 44.4933
LR_7[cache_size=0.2,treshold=0.7] 30.2947 30.2775 30.2891 30.2932
LR_7[cache_size=0.2,treshold=0.8] 10.4374 10.3694 10.4005 10.3963
LR_7[cache_size=0.2,treshold=0.9] 0.00777916 0.00636331 0.00716753 0.00743969
LR_8[cache_size=0.2,treshold=0.3] 69.7063 69.6769 69.6931 69.6982
LR_8[cache_size=0.2,treshold=0.5] 54.9255 54.8923 54.9038 54.9009
LR_8[cache_size=0.2,treshold=0.6] 44.5447 44.5027 44.5249 44.5246
LR_8[cache_size=0.2,treshold=0.7] 30.2211 30.2054 30.2151 30.2175
LR_8[cache_size=0.2,treshold=0.8] 10.2106 10.1508 10.1785 10.1771
LR_8[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.2,treshold=0.3] 69.7393 69.7092 69.7248 69.7293
LR_9[cache_size=0.2,treshold=0.5] 54.952 54.9204 54.9308 54.9263
LR_9[cache_size=0.2,treshold=0.6] 44.5602 44.5161 44.5394 44.5399
LR_9[cache_size=0.2,treshold=0.7] 30.2082 30.1953 30.2029 30.2049
LR_9[cache_size=0.2,treshold=0.8] 10.158 10.0991 10.1267 10.1253
LR_9[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_10[cache_size=0.4,treshold=0.3] 74.7041 74.6631 74.6901 74.6924
LR_10[cache_size=0.4,treshold=0.5] 56.2596 56.2169 56.2384 56.2369
LR_10[cache_size=0.4,treshold=0.6] 42.7017 42.5409 42.6008 42.5851
LR_10[cache_size=0.4,treshold=0.7] 21.6635 21.515 21.5831 21.5802
LR_10[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_10[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.4,treshold=0.3] 74.6961 74.6552 74.6818 74.6839
LR_11[cache_size=0.4,treshold=0.5] 56.2589 56.2152 56.2375 56.2363
LR_11[cache_size=0.4,treshold=0.6] 42.7111 42.5457 42.6074 42.5912
LR_11[cache_size=0.4,treshold=0.7] 21.6802 21.5303 21.5984 21.5947
LR_11[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.4,treshold=0.3] 74.7116 74.6719 74.6982 74.7006
LR_12[cache_size=0.4,treshold=0.5] 56.2644 56.2215 56.2429 56.2407
LR_12[cache_size=0.4,treshold=0.6] 42.701 42.5402 42.5999 42.584
LR_12[cache_size=0.4,treshold=0.7] 21.6484 21.4979 21.5674 21.5647
LR_12[cache_size=0.4,treshold=0.8] 6.49243e-05 0 1.94808e-05 1.08217e-05
LR_12[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.3] 97.245 97.2333 97.2396 97.239
LR_13[cache_size=0.4,treshold=0.5] 85.5184 85.4949 85.5064 85.5049
LR_13[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.3] 97.245 97.2333 97.2396 97.239
LR_14[cache_size=0.4,treshold=0.5] 85.5184 85.4949 85.5064 85.5049
LR_14[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.3] 97.245 97.2333 97.2396 97.239
LR_15[cache_size=0.4,treshold=0.5] 85.5184 85.4949 85.5064 85.5049
LR_15[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_7[cache_size=0.4,treshold=0.3] 71.2408 71.1857 71.2103 71.2129
LR_7[cache_size=0.4,treshold=0.5] 56.0155 55.9463 55.9762 55.9707
LR_7[cache_size=0.4,treshold=0.6] 44.7207 44.6566 44.6876 44.6848
LR_7[cache_size=0.4,treshold=0.7] 28.1993 28.0878 28.1311 28.1207
LR_7[cache_size=0.4,treshold=0.8] 6.74312 6.65097 6.69731 6.71616
LR_7[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_8[cache_size=0.4,treshold=0.3] 71.2314 71.1766 71.201 71.2032
LR_8[cache_size=0.4,treshold=0.5] 56.0076 55.939 55.9691 55.964
LR_8[cache_size=0.4,treshold=0.6] 44.7167 44.653 44.6843 44.6826
LR_8[cache_size=0.4,treshold=0.7] 28.2041 28.0924 28.1352 28.1246
LR_8[cache_size=0.4,treshold=0.8] 6.75206 6.66728 6.7124 6.72824
LR_8[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.4,treshold=0.3] 71.2425 71.1869 71.2115 71.2141
LR_9[cache_size=0.4,treshold=0.5] 56.0155 55.947 55.9765 55.9711
LR_9[cache_size=0.4,treshold=0.6] 44.718 44.654 44.6854 44.6834
LR_9[cache_size=0.4,treshold=0.7] 28.1928 28.081 28.1242 28.114
LR_9[cache_size=0.4,treshold=0.8] 6.73749 6.63896 6.68599 6.69981
LR_9[cache_size=0.4,treshold=0.9] 0 0 0 0
Offline Clock 1st iteration 0 0 0 0
Offline Clock 2nd iteration 43.4673 41.4753 42.6567 43.0175
Zipf Optimal Distribution 17.0869 6.59027 12.1282 13.1121

Miss Ratio Reduced (%)

Model Max Min Avg Mdn
LR_10[cache_size=0.001,treshold=0.3] -3.21687 -3.2229 -3.2198 -3.21918
LR_10[cache_size=0.001,treshold=0.5] -2.17022 -2.17461 -2.17256 -2.17253
LR_10[cache_size=0.001,treshold=0.6] -1.58558 -1.591 -1.58861 -1.58953
LR_10[cache_size=0.001,treshold=0.7] -0.931972 -0.93897 -0.936225 -0.936453
LR_10[cache_size=0.001,treshold=0.8] -0.267959 -0.269508 -0.268521 -0.268143
LR_10[cache_size=0.001,treshold=0.9] 0.000682298 0.000341196 0.000511724 0.000511706
LR_10[cache_size=All,treshold=0.3] -3.25695 -26.6916 -15.5345 -17.181
LR_10[cache_size=All,treshold=0.5] -0.0279728 -26.4195 -13.7829 -16.2085
LR_10[cache_size=All,treshold=0.6] 0 0 0 0
LR_10[cache_size=All,treshold=0.7] 0 0 0 0
LR_10[cache_size=All,treshold=0.8] 0 0 0 0
LR_10[cache_size=All,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.001,treshold=0.3] -3.2312 -3.23603 -3.23314 -3.23256
LR_11[cache_size=0.001,treshold=0.5] -2.15299 -2.15687 -2.15499 -2.15514
LR_11[cache_size=0.001,treshold=0.6] -1.5559 -1.56165 -1.5592 -1.56036
LR_11[cache_size=0.001,treshold=0.7] -0.908775 -0.914916 -0.912481 -0.912402
LR_11[cache_size=0.001,treshold=0.8] -0.270227 -0.272228 -0.271284 -0.27137
LR_11[cache_size=0.001,treshold=0.9] 0.00818756 0.0075049 0.00764181 0.00750528
LR_11[cache_size=All,treshold=0.3] -3.25661 -26.6916 -15.5343 -17.181
LR_11[cache_size=All,treshold=0.5] -0.0286565 -26.4207 -13.787 -16.2117
LR_11[cache_size=All,treshold=0.6] 0 0 0 0
LR_11[cache_size=All,treshold=0.7] 0 0 0 0
LR_11[cache_size=All,treshold=0.8] 0 0 0 0
LR_11[cache_size=All,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.001,treshold=0.3] -3.21499 -3.22051 -3.21748 -3.21696
LR_12[cache_size=0.001,treshold=0.5] -2.16186 -2.16557 -2.16376 -2.16383
LR_12[cache_size=0.001,treshold=0.6] -1.57944 -1.58502 -1.58291 -1.58407
LR_12[cache_size=0.001,treshold=0.7] -0.936577 -0.943235 -0.940626 -0.940546
LR_12[cache_size=0.001,treshold=0.8] -0.286906 -0.288953 -0.287694 -0.287287
LR_12[cache_size=0.001,treshold=0.9] 0.00579952 0.00511724 0.00539021 0.00528853
LR_12[cache_size=All,treshold=0.3] -3.25593 -26.6927 -15.5354 -17.182
LR_12[cache_size=All,treshold=0.5] -0.0267802 -26.4184 -13.7803 -16.2071
LR_12[cache_size=All,treshold=0.6] 0 0 0 0
LR_12[cache_size=All,treshold=0.7] 0 0 0 0
LR_12[cache_size=All,treshold=0.8] 0 0 0 0
LR_12[cache_size=All,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.001,treshold=0.3] -3.52662 -3.53197 -3.52902 -3.5285
LR_13[cache_size=0.001,treshold=0.5] -2.43687 -2.44023 -2.43873 -2.43887
LR_13[cache_size=0.001,treshold=0.6] -1.80015 -1.80615 -1.80436 -1.80502
LR_13[cache_size=0.001,treshold=0.7] -1.06348 -1.06914 -1.06644 -1.06626
LR_13[cache_size=0.001,treshold=0.8] -0.281945 -0.284859 -0.283736 -0.283665
LR_13[cache_size=0.001,treshold=0.9] 0 0 0 0
LR_13[cache_size=All,treshold=0.3] -3.55613 -26.7584 -15.8329 -17.4288
LR_13[cache_size=All,treshold=0.5] 0.00136459 -26.5366 -14.077 -16.7173
LR_13[cache_size=All,treshold=0.6] 0 0 0 0
LR_13[cache_size=All,treshold=0.7] 0 0 0 0
LR_13[cache_size=All,treshold=0.8] 0 0 0 0
LR_13[cache_size=All,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.001,treshold=0.3] -3.54368 -3.54834 -3.54587 -3.54556
LR_14[cache_size=0.001,treshold=0.5] -2.40702 -2.41116 -2.40918 -2.40851
LR_14[cache_size=0.001,treshold=0.6] -1.75393 -1.76057 -1.75864 -1.75914
LR_14[cache_size=0.001,treshold=0.7] -1.02288 -1.02785 -1.02561 -1.02532
LR_14[cache_size=0.001,treshold=0.8] -0.296102 -0.299007 -0.297553 -0.297311
LR_14[cache_size=0.001,treshold=0.9] 0 0 0 0
LR_14[cache_size=All,treshold=0.3] -3.57318 -26.7573 -15.8342 -17.4278
LR_14[cache_size=All,treshold=0.5] 0.00221747 -26.5199 -13.9948 -16.6568
LR_14[cache_size=All,treshold=0.6] 0 0 0 0
LR_14[cache_size=All,treshold=0.7] 0 0 0 0
LR_14[cache_size=All,treshold=0.8] 0 0 0 0
LR_14[cache_size=All,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.001,treshold=0.3] -3.52662 -3.53094 -3.52806 -3.52765
LR_15[cache_size=0.001,treshold=0.5] -2.42476 -2.42822 -2.42645 -2.42625
LR_15[cache_size=0.001,treshold=0.6] -1.78804 -1.79469 -1.79289 -1.79376
LR_15[cache_size=0.001,treshold=0.7] -1.06467 -1.06965 -1.06736 -1.06711
LR_15[cache_size=0.001,treshold=0.8] -0.304289 -0.307546 -0.306048 -0.305541
LR_15[cache_size=0.001,treshold=0.9] 0 0 0 0
LR_15[cache_size=All,treshold=0.3] -3.56738 -26.7584 -15.8349 -17.4288
LR_15[cache_size=All,treshold=0.5] 0.00238804 -26.5154 -13.9961 -16.6522
LR_15[cache_size=All,treshold=0.6] 0 0 0 0
LR_15[cache_size=All,treshold=0.7] 0 0 0 0
LR_15[cache_size=All,treshold=0.8] 0 0 0 0
LR_15[cache_size=All,treshold=0.9] 0 0 0 0
LR_7[cache_size=0.001,treshold=0.3] -3.97043 -3.97459 -3.97269 -3.97284
LR_7[cache_size=0.001,treshold=0.5] -2.68215 -2.6876 -2.68456 -2.68416
LR_7[cache_size=0.001,treshold=0.6] -1.90788 -1.91206 -1.91004 -1.90958
LR_7[cache_size=0.001,treshold=0.7] -1.00809 -1.01181 -1.00964 -1.00895
LR_7[cache_size=0.001,treshold=0.8] -0.118736 -0.120596 -0.119813 -0.119908
LR_7[cache_size=0.001,treshold=0.9] 0.0279742 0.0264382 0.0270193 0.0269494
LR_7[cache_size=All,treshold=0.3] -3.70077 -26.0015 -15.5442 -17.1022
LR_7[cache_size=All,treshold=0.5] -0.231799 -23.5434 -12.9866 -15.4492
LR_7[cache_size=All,treshold=0.6] 0 0 0 0
LR_7[cache_size=All,treshold=0.7] 0 0 0 0
LR_7[cache_size=All,treshold=0.8] 0 0 0 0
LR_7[cache_size=All,treshold=0.9] 0 0 0 0
LR_8[cache_size=0.001,treshold=0.3] -4.01444 -4.01775 -4.01608 -4.01587
LR_8[cache_size=0.001,treshold=0.5] -2.63729 -2.64109 -2.63888 -2.63759
LR_8[cache_size=0.001,treshold=0.6] -1.83384 -1.83685 -1.83502 -1.83453
LR_8[cache_size=0.001,treshold=0.7] -0.98609 -0.990322 -0.98825 -0.988138
LR_8[cache_size=0.001,treshold=0.8] -0.180322 -0.182173 -0.181459 -0.181491
LR_8[cache_size=0.001,treshold=0.9] 0.0262684 0.024732 0.0251088 0.0247333
LR_8[cache_size=All,treshold=0.3] -3.76882 -25.9926 -15.5579 -17.0987
LR_8[cache_size=All,treshold=0.5] -0.198197 -23.433 -12.7635 -15.23
LR_8[cache_size=All,treshold=0.6] 0 0 0 0
LR_8[cache_size=All,treshold=0.7] 0 0 0 0
LR_8[cache_size=All,treshold=0.8] 0 0 0 0
LR_8[cache_size=All,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.001,treshold=0.3] -3.96548 -3.96965 -3.96764 -3.96773
LR_9[cache_size=0.001,treshold=0.5] -2.66181 -2.66627 -2.66403 -2.66403
LR_9[cache_size=0.001,treshold=0.6] -1.89389 -1.89739 -1.89551 -1.89482
LR_9[cache_size=0.001,treshold=0.7] -1.01543 -1.0183 -1.01657 -1.0164
LR_9[cache_size=0.001,treshold=0.8] -0.148591 -0.150271 -0.149664 -0.149935
LR_9[cache_size=0.001,treshold=0.9] 0.0279742 0.0262676 0.0268828 0.0267788
LR_9[cache_size=All,treshold=0.3] -3.75808 -25.997 -15.5573 -17.1001
LR_9[cache_size=All,treshold=0.5] -0.170939 -23.4665 -12.7712 -15.2422
LR_9[cache_size=All,treshold=0.6] 0 0 0 0
LR_9[cache_size=All,treshold=0.7] 0 0 0 0
LR_9[cache_size=All,treshold=0.8] 0 0 0 0
LR_9[cache_size=All,treshold=0.9] 0 0 0 0
LR_10[cache_size=0.01,treshold=0.3] -4.75913 -4.77055 -4.76281 -4.76185
LR_10[cache_size=0.01,treshold=0.5] -3.23771 -3.25293 -3.24348 -3.24177
LR_10[cache_size=0.01,treshold=0.6] -2.36641 -2.37553 -2.36954 -2.36871
LR_10[cache_size=0.01,treshold=0.7] -1.37419 -1.38484 -1.37991 -1.37992
LR_10[cache_size=0.01,treshold=0.8] -0.381734 -0.390036 -0.385496 -0.386683
LR_10[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.01,treshold=0.3] -4.75617 -4.76784 -4.75966 -4.75839
LR_11[cache_size=0.01,treshold=0.5] -3.22981 -3.24479 -3.23563 -3.23412
LR_11[cache_size=0.01,treshold=0.6] -2.36155 -2.37109 -2.36485 -2.36402
LR_11[cache_size=0.01,treshold=0.7] -1.37715 -1.38731 -1.38218 -1.38214
LR_11[cache_size=0.01,treshold=0.8] -0.390178 -0.398917 -0.394232 -0.395566
LR_11[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.01,treshold=0.3] -4.76061 -4.77179 -4.7643 -4.76308
LR_12[cache_size=0.01,treshold=0.5] -3.24116 -3.25713 -3.24723 -3.24537
LR_12[cache_size=0.01,treshold=0.6] -2.36789 -2.37701 -2.37087 -2.3697
LR_12[cache_size=0.01,treshold=0.7] -1.36926 -1.3804 -1.37498 -1.37473
LR_12[cache_size=0.01,treshold=0.8] -0.371864 -0.380908 -0.37607 -0.377059
LR_12[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.01,treshold=0.3] -5.92757 -5.94018 -5.93145 -5.93065
LR_13[cache_size=0.01,treshold=0.5] -4.06459 -4.07702 -4.07 -4.06935
LR_13[cache_size=0.01,treshold=0.6] -2.94703 -2.95874 -2.9525 -2.95253
LR_13[cache_size=0.01,treshold=0.7] -1.70312 -1.71507 -1.71003 -1.71132
LR_13[cache_size=0.01,treshold=0.8] -0.293395 -0.296783 -0.295378 -0.295903
LR_13[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.01,treshold=0.3] -5.91749 -5.93105 -5.92187 -5.92127
LR_14[cache_size=0.01,treshold=0.5] -4.04411 -4.05629 -4.04956 -4.04936
LR_14[cache_size=0.01,treshold=0.6] -2.93617 -2.94812 -2.94184 -2.94217
LR_14[cache_size=0.01,treshold=0.7] -1.7051 -1.71729 -1.71196 -1.7128
LR_14[cache_size=0.01,treshold=0.8] -0.316343 -0.319617 -0.318179 -0.318361
LR_14[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.01,treshold=0.3] -5.93102 -5.94314 -5.93495 -5.93411
LR_15[cache_size=0.01,treshold=0.5] -4.07101 -4.08369 -4.07661 -4.07577
LR_15[cache_size=0.01,treshold=0.6] -2.95048 -2.96194 -2.95541 -2.95525
LR_15[cache_size=0.01,treshold=0.7] -1.70164 -1.71384 -1.70865 -1.71034
LR_15[cache_size=0.01,treshold=0.8] -0.285005 -0.288642 -0.287136 -0.287759
LR_15[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_7[cache_size=0.01,treshold=0.3] -5.19672 -5.21209 -5.20132 -5.19862
LR_7[cache_size=0.01,treshold=0.5] -3.43931 -3.45556 -3.44835 -3.44979
LR_7[cache_size=0.01,treshold=0.6] -2.44611 -2.46043 -2.45378 -2.45409
LR_7[cache_size=0.01,treshold=0.7] -1.33841 -1.34486 -1.34122 -1.34167
LR_7[cache_size=0.01,treshold=0.8] -0.234699 -0.237081 -0.235809 -0.235455
LR_7[cache_size=0.01,treshold=0.9] 0.0399803 0.0380085 0.0390876 0.0389789
LR_8[cache_size=0.01,treshold=0.3] -5.19474 -5.21061 -5.19949 -5.19665
LR_8[cache_size=0.01,treshold=0.5] -3.43684 -3.45334 -3.44608 -3.44757
LR_8[cache_size=0.01,treshold=0.6] -2.44537 -2.45969 -2.45309 -2.4536
LR_8[cache_size=0.01,treshold=0.7] -1.34039 -1.34659 -1.34314 -1.34364
LR_8[cache_size=0.01,treshold=0.8] -0.238368 -0.241028 -0.239461 -0.239157
LR_8[cache_size=0.01,treshold=0.9] 0.0399761 0.0377616 0.0388902 0.0387322
LR_9[cache_size=0.01,treshold=0.3] -5.2172 -5.23283 -5.22145 -5.21849
LR_9[cache_size=0.01,treshold=0.5] -3.45781 -3.47284 -3.46606 -3.46706
LR_9[cache_size=0.01,treshold=0.6] -2.45006 -2.46413 -2.45739 -2.4568
LR_9[cache_size=0.01,treshold=0.7] -1.31917 -1.32585 -1.32227 -1.32267
LR_9[cache_size=0.01,treshold=0.8] -0.20311 -0.20649 -0.205012 -0.205344
LR_9[cache_size=0.01,treshold=0.9] 0.0402229 0.0380085 0.0391863 0.0389789
LR_10[cache_size=0.1,treshold=0.3] -10.324 -10.3411 -10.3285 -10.3258
LR_10[cache_size=0.1,treshold=0.5] -7.98478 -7.99936 -7.9908 -7.98939
LR_10[cache_size=0.1,treshold=0.6] -6.21124 -6.22448 -6.21908 -6.22072
LR_10[cache_size=0.1,treshold=0.7] -2.23289 -2.24142 -2.23817 -2.2393
LR_10[cache_size=0.1,treshold=0.8] -0.00187072 -0.00233845 -0.00224432 -0.00233774
LR_10[cache_size=0.1,treshold=0.9] -0.00140198 -0.00140307 -0.00140271 -0.00140284
LR_11[cache_size=0.1,treshold=0.3] -10.3349 -10.3518 -10.3396 -10.3375
LR_11[cache_size=0.1,treshold=0.5] -8.03948 -8.05502 -8.04579 -8.04363
LR_11[cache_size=0.1,treshold=0.6] -6.25797 -6.27172 -6.26686 -6.27028
LR_11[cache_size=0.1,treshold=0.7] -2.13496 -2.14277 -2.13811 -2.13781
LR_11[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.1,treshold=0.3] -10.3333 -10.3504 -10.3382 -10.3361
LR_12[cache_size=0.1,treshold=0.5] -8.07314 -8.08776 -8.0787 -8.07728
LR_12[cache_size=0.1,treshold=0.6] -6.28648 -6.30072 -6.29576 -6.29833
LR_12[cache_size=0.1,treshold=0.7] -2.09474 -2.10396 -2.09902 -2.09899
LR_12[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_12[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.1,treshold=0.3] -10.6382 -10.6558 -10.6443 -10.6421
LR_13[cache_size=0.1,treshold=0.5] -7.10103 -7.1159 -7.10896 -7.10953
LR_13[cache_size=0.1,treshold=0.6] -5.02797 -5.04321 -5.0378 -5.03969
LR_13[cache_size=0.1,treshold=0.7] -2.5044 -2.51482 -2.50955 -2.51009
LR_13[cache_size=0.1,treshold=0.8] -0.000467548 -0.000935362 -0.000748096 -0.000934654
LR_13[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.1,treshold=0.3] -10.6733 -10.6895 -10.6785 -10.6762
LR_14[cache_size=0.1,treshold=0.5] -7.11225 -7.12619 -7.12009 -7.12075
LR_14[cache_size=0.1,treshold=0.6] -5.0275 -5.04321 -5.03752 -5.03969
LR_14[cache_size=0.1,treshold=0.7] -2.47683 -2.48583 -2.48103 -2.48063
LR_14[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_14[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.1,treshold=0.3] -10.6836 -10.7003 -10.6889 -10.686
LR_15[cache_size=0.1,treshold=0.5] -7.1244 -7.13741 -7.1315 -7.13244
LR_15[cache_size=0.1,treshold=0.6] -5.0247 -5.03993 -5.03453 -5.03689
LR_15[cache_size=0.1,treshold=0.7] -2.46515 -2.47274 -2.46878 -2.46847
LR_15[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_15[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_7[cache_size=0.1,treshold=0.3] -7.48836 -7.50548 -7.49648 -7.49505
LR_7[cache_size=0.1,treshold=0.5] -4.77549 -4.79008 -4.78344 -4.78262
LR_7[cache_size=0.1,treshold=0.6] -3.2934 -3.30703 -3.2989 -3.29839
LR_7[cache_size=0.1,treshold=0.7] -1.69359 -1.70679 -1.70037 -1.70187
LR_7[cache_size=0.1,treshold=0.8] -0.231974 -0.239418 -0.235936 -0.236579
LR_7[cache_size=0.1,treshold=0.9] -0.00140198 -0.00140307 -0.00140271 -0.00140284
LR_8[cache_size=0.1,treshold=0.3] -7.5122 -7.52793 -7.51883 -7.51797
LR_8[cache_size=0.1,treshold=0.5] -4.78204 -4.79709 -4.78999 -4.78862
LR_8[cache_size=0.1,treshold=0.6] -3.28639 -3.30095 -3.29227 -3.29092
LR_8[cache_size=0.1,treshold=0.7] -1.67256 -1.68575 -1.67877 -1.6799
LR_8[cache_size=0.1,treshold=0.8] -0.209525 -0.216505 -0.213212 -0.213101
LR_8[cache_size=0.1,treshold=0.9] 0 -0.000467681 -0.000280511 -0.000467327
LR_9[cache_size=0.1,treshold=0.3] -7.51501 -7.53074 -7.52154 -7.52078
LR_9[cache_size=0.1,treshold=0.5] -4.78297 -4.7985 -4.79111 -4.78917
LR_9[cache_size=0.1,treshold=0.6] -3.28546 -3.30049 -3.2917 -3.29045
LR_9[cache_size=0.1,treshold=0.7] -1.66976 -1.68341 -1.67615 -1.67714
LR_9[cache_size=0.1,treshold=0.8] -0.205783 -0.212734 -0.209565 -0.209362
LR_9[cache_size=0.1,treshold=0.9] 0 -0.000467681 -0.000280511 -0.000467327
LR_10[cache_size=0.2,treshold=0.3] -12.2754 -12.2815 -12.2776 -12.2778
LR_10[cache_size=0.2,treshold=0.5] -8.05348 -8.06355 -8.0599 -8.06154
LR_10[cache_size=0.2,treshold=0.6] -5.64452 -5.66228 -5.6566 -5.65871
LR_10[cache_size=0.2,treshold=0.7] -2.75628 -2.76833 -2.7616 -2.76271
LR_10[cache_size=0.2,treshold=0.8] -0.00196289 -0.00196456 -0.00196369 -0.00196354
LR_10[cache_size=0.2,treshold=0.9] -0.00130859 -0.00130971 -0.00130913 -0.00130903
LR_11[cache_size=0.2,treshold=0.3] -12.292 -12.2985 -12.2953 -12.2955
LR_11[cache_size=0.2,treshold=0.5] -8.05938 -8.07009 -8.06579 -8.06743
LR_11[cache_size=0.2,treshold=0.6] -5.64124 -5.65966 -5.65346 -5.65544
LR_11[cache_size=0.2,treshold=0.7] -2.73663 -2.75001 -2.74236 -2.74307
LR_11[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.2,treshold=0.3] -12.3018 -12.3083 -12.3045 -12.304
LR_12[cache_size=0.2,treshold=0.5] -8.06265 -8.07271 -8.0688 -8.07041
LR_12[cache_size=0.2,treshold=0.6] -5.63928 -5.6577 -5.6515 -5.65335
LR_12[cache_size=0.2,treshold=0.7] -2.72922 -2.74215 -2.73463 -2.73522
LR_12[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_12[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.2,treshold=0.3] -12.2978 -12.3044 -12.3007 -12.3008
LR_13[cache_size=0.2,treshold=0.5] -8.06068 -8.0714 -8.06723 -8.0691
LR_13[cache_size=0.2,treshold=0.6] -5.64059 -5.65901 -5.65255 -5.65466
LR_13[cache_size=0.2,treshold=0.7] -2.73336 -2.74608 -2.73908 -2.7398
LR_13[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_13[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.2,treshold=0.3] -12.292 -12.2985 -12.2954 -12.2955
LR_14[cache_size=0.2,treshold=0.5] -8.05938 -8.07009 -8.06592 -8.06779
LR_14[cache_size=0.2,treshold=0.6] -5.64124 -5.65966 -5.65346 -5.65544
LR_14[cache_size=0.2,treshold=0.7] -2.73663 -2.75001 -2.74236 -2.74307
LR_14[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_14[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.2,treshold=0.3] -12.3011 -12.3083 -12.3044 -12.304
LR_15[cache_size=0.2,treshold=0.5] -8.06265 -8.07336 -8.06933 -8.07106
LR_15[cache_size=0.2,treshold=0.6] -5.63993 -5.65835 -5.65202 -5.654
LR_15[cache_size=0.2,treshold=0.7] -2.72922 -2.74215 -2.73463 -2.73522
LR_15[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_15[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_7[cache_size=0.2,treshold=0.3] -8.57414 -8.57858 -8.57596 -8.57545
LR_7[cache_size=0.2,treshold=0.5] -5.22041 -5.23591 -5.22656 -5.22357
LR_7[cache_size=0.2,treshold=0.6] -3.42769 -3.44523 -3.43567 -3.43798
LR_7[cache_size=0.2,treshold=0.7] -1.56102 -1.57424 -1.56833 -1.56881
LR_7[cache_size=0.2,treshold=0.8] -0.0824656 -0.0942717 -0.0864022 -0.0844324
LR_7[cache_size=0.2,treshold=0.9] -0.00196289 -0.00261941 -0.00209466 -0.00196354
LR_8[cache_size=0.2,treshold=0.3] -8.59705 -8.60215 -8.599 -8.59867
LR_8[cache_size=0.2,treshold=0.5] -5.2263 -5.24376 -5.23349 -5.23031
LR_8[cache_size=0.2,treshold=0.6] -3.42246 -3.43868 -3.42978 -3.43143
LR_8[cache_size=0.2,treshold=0.7] -1.54465 -1.55526 -1.55053 -1.55114
LR_8[cache_size=0.2,treshold=0.8] -0.0687596 -0.0792144 -0.0729181 -0.0719966
LR_8[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.2,treshold=0.3] -8.60164 -8.60739 -8.6041 -8.60359
LR_9[cache_size=0.2,treshold=0.5] -5.22892 -5.24507 -5.23546 -5.23359
LR_9[cache_size=0.2,treshold=0.6] -3.4218 -3.43868 -3.42939 -3.43078
LR_9[cache_size=0.2,treshold=0.7] -1.54138 -1.55264 -1.54791 -1.54852
LR_9[cache_size=0.2,treshold=0.8] -0.0674499 -0.0779051 -0.0718708 -0.0713421
LR_9[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_10[cache_size=0.4,treshold=0.3] -14.2156 -14.275 -14.2511 -14.2538
LR_10[cache_size=0.4,treshold=0.5] -8.93748 -8.98745 -8.96031 -8.95992
LR_10[cache_size=0.4,treshold=0.6] -6.01236 -6.06569 -6.04422 -6.04908
LR_10[cache_size=0.4,treshold=0.7] -2.61501 -2.64685 -2.63335 -2.64157
LR_10[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_10[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.4,treshold=0.3] -14.2123 -14.2728 -14.2484 -14.2516
LR_11[cache_size=0.4,treshold=0.5] -8.93726 -8.98711 -8.96007 -8.95981
LR_11[cache_size=0.4,treshold=0.6] -6.01314 -6.06703 -6.0452 -6.05019
LR_11[cache_size=0.4,treshold=0.7] -2.61635 -2.64852 -2.63487 -2.64336
LR_11[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.4,treshold=0.3] -14.2179 -14.2784 -14.254 -14.2571
LR_12[cache_size=0.4,treshold=0.5] -8.93815 -8.98867 -8.96125 -8.96093
LR_12[cache_size=0.4,treshold=0.6] -6.01203 -6.06558 -6.0441 -6.0493
LR_12[cache_size=0.4,treshold=0.7] -2.61367 -2.64473 -2.63128 -2.63934
LR_12[cache_size=0.4,treshold=0.8] 0 -0.000111479 -2.22958e-05 0
LR_12[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.3] -24.7111 -24.7763 -24.7426 -24.7333
LR_13[cache_size=0.4,treshold=0.5] -18.4332 -18.4889 -18.4702 -18.4734
LR_13[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.3] -24.7111 -24.7763 -24.7426 -24.7333
LR_14[cache_size=0.4,treshold=0.5] -18.4332 -18.4889 -18.4702 -18.4734
LR_14[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.3] -24.7111 -24.7763 -24.7426 -24.7333
LR_15[cache_size=0.4,treshold=0.5] -18.4332 -18.4889 -18.4702 -18.4734
LR_15[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_7[cache_size=0.4,treshold=0.3] -9.46621 -9.52188 -9.49432 -9.49426
LR_7[cache_size=0.4,treshold=0.5] -5.35293 -5.42246 -5.39159 -5.38726
LR_7[cache_size=0.4,treshold=0.6] -3.23244 -3.29443 -3.26094 -3.25586
LR_7[cache_size=0.4,treshold=0.7] -1.19619 -1.23474 -1.22055 -1.22163
LR_7[cache_size=0.4,treshold=0.8] -0.102515 -0.129588 -0.117868 -0.122509
LR_7[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_8[cache_size=0.4,treshold=0.3] -9.46543 -9.52076 -9.4928 -9.49181
LR_8[cache_size=0.4,treshold=0.5] -5.35315 -5.42279 -5.39183 -5.38793
LR_8[cache_size=0.4,treshold=0.6] -3.23378 -3.29477 -3.26206 -3.25686
LR_8[cache_size=0.4,treshold=0.7] -1.19797 -1.2363 -1.22197 -1.2223
LR_8[cache_size=0.4,treshold=0.8] -0.103295 -0.130034 -0.118113 -0.122286
LR_8[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.4,treshold=0.3] -9.46666 -9.52165 -9.49456 -9.49459
LR_9[cache_size=0.4,treshold=0.5] -5.35326 -5.42312 -5.39172 -5.38693
LR_9[cache_size=0.4,treshold=0.6] -3.23233 -3.29443 -3.2609 -3.25575
LR_9[cache_size=0.4,treshold=0.7] -1.19608 -1.23507 -1.22055 -1.2213
LR_9[cache_size=0.4,treshold=0.8] -0.103183 -0.129477 -0.118047 -0.122509
LR_9[cache_size=0.4,treshold=0.9] 0 0 0 0
Offline Clock 1st iteration 0 0 0 0
Offline Clock 2nd iteration 7.80369 0.100977 2.73498 1.84941
Zipf Optimal Distribution 3.35401 0.0684004 1.23463 0.902408

Model Summaries Plot

Miss Ratio Reduced (%)

Promotion Reduced (%)

Promotion vs Miss Ratio

Cache Size All

Cache Size 0.001

Cache Size 0.01

Cache Size 0.1

Cache Size 0.2

Cache Size 0.4

Model Classification Report

LR_10_0.001

              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.87      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.78      0.79      0.77  50357831
weighted avg       0.80      0.77      0.78  50357831

Accuracy: 0.7738074342399696
              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.87      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.78      0.79      0.77  50357831
weighted avg       0.80      0.77      0.78  50357831

Accuracy: 0.7738074342399696

LR_10_0.01

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7611738202303706
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7611738202303706

LR_10_0.1

              precision    recall  f1-score   support

           0       0.87      0.59      0.70  29263977
           1       0.58      0.86      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.72      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.698262447511934
              precision    recall  f1-score   support

           0       0.87      0.59      0.70  29263977
           1       0.58      0.86      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.72      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.698262447511934

LR_10_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202710005355438
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202710005355438

LR_10_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7429382331601068
              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7429382331601068

LR_10_All

              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.5060794899989584
              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.5060794899989584

LR_11_0.001

              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.86      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.77      0.79      0.77  50357831
weighted avg       0.80      0.77      0.77  50357831

Accuracy: 0.7689939822864889
              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.86      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.77      0.79      0.77  50357831
weighted avg       0.80      0.77      0.77  50357831

Accuracy: 0.7689939822864889

LR_11_0.01

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7614748322626786
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7614748322626786

LR_11_0.1

              precision    recall  f1-score   support

           0       0.87      0.59      0.70  29263977
           1       0.58      0.87      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.73      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.6982573486401464
              precision    recall  f1-score   support

           0       0.87      0.59      0.70  29263977
           1       0.58      0.87      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.73      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.6982573486401464

LR_11_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202601258885529
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202601258885529

LR_11_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.742940891443532
              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.742940891443532

LR_11_All

              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.5060570082220569
              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.5060570082220569

LR_12_0.001

              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.87      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.78      0.79      0.77  50357831
weighted avg       0.80      0.77      0.78  50357831

Accuracy: 0.7734336691347965
              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.87      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.78      0.79      0.77  50357831
weighted avg       0.80      0.77      0.78  50357831

Accuracy: 0.7734336691347965

LR_12_0.01

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7608722086919237
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7608722086919237

LR_12_0.1

              precision    recall  f1-score   support

           0       0.88      0.59      0.70  29263977
           1       0.58      0.87      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.73      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.6982929785612961
              precision    recall  f1-score   support

           0       0.88      0.59      0.70  29263977
           1       0.58      0.87      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.73      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.6982929785612961

LR_12_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202365352698812
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202365352698812

LR_12_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7429330642756687
              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7429330642756687

LR_12_All

              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.506046695402424
              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.506046695402424

LR_13_0.001

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30469299
           1       0.65      0.87      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.77      0.78      0.76  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7611932094533619
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30469299
           1       0.65      0.87      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.77      0.78      0.76  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7611932094533619

LR_13_0.01

              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379024507191466
              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379024507191466

LR_13_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7041242431724914
              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7041242431724914

LR_13_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202479514829574
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202479514829574

LR_13_0.4

              precision    recall  f1-score   support

           0       0.89      0.66      0.76  26121772
           1       0.58      0.86      0.70  14505950

    accuracy                           0.73  40627722
   macro avg       0.74      0.76      0.73  40627722
weighted avg       0.78      0.73      0.74  40627722

Accuracy: 0.7319589811114686
              precision    recall  f1-score   support

           0       0.89      0.66      0.76  26121772
           1       0.58      0.86      0.70  14505950

    accuracy                           0.73  40627722
   macro avg       0.74      0.76      0.73  40627722
weighted avg       0.78      0.73      0.74  40627722

Accuracy: 0.7319589811114686

LR_13_All

              precision    recall  f1-score   support

           0       0.69      0.31      0.43 144486817
           1       0.41      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.54      0.49 235435127
weighted avg       0.58      0.49      0.47 235435127

Accuracy: 0.49110166555562457
              precision    recall  f1-score   support

           0       0.69      0.31      0.43 144486817
           1       0.41      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.54      0.49 235435127
weighted avg       0.58      0.49      0.47 235435127

Accuracy: 0.49110166555562457

LR_14_0.001

              precision    recall  f1-score   support

           0       0.88      0.69      0.77  30469299
           1       0.64      0.86      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.76      0.77      0.75  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7556111779317898
              precision    recall  f1-score   support

           0       0.88      0.69      0.77  30469299
           1       0.64      0.86      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.76      0.77      0.75  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7556111779317898

LR_14_0.01

              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379092651055913
              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379092651055913

LR_14_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7040489622687012
              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7040489622687012

LR_14_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202581329253593
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202581329253593

LR_14_0.4

              precision    recall  f1-score   support

           0       0.90      0.63      0.74  26121772
           1       0.57      0.87      0.69  14505950

    accuracy                           0.72  40627722
   macro avg       0.73      0.75      0.71  40627722
weighted avg       0.78      0.72      0.72  40627722

Accuracy: 0.7152666595483744
              precision    recall  f1-score   support

           0       0.90      0.63      0.74  26121772
           1       0.57      0.87      0.69  14505950

    accuracy                           0.72  40627722
   macro avg       0.73      0.75      0.71  40627722
weighted avg       0.78      0.72      0.72  40627722

Accuracy: 0.7152666595483744

LR_14_All

              precision    recall  f1-score   support

           0       0.69      0.32      0.44 144486817
           1       0.42      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.55      0.49 235435127
weighted avg       0.58      0.49      0.48 235435127

Accuracy: 0.4946263604920773
              precision    recall  f1-score   support

           0       0.69      0.32      0.44 144486817
           1       0.42      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.55      0.49 235435127
weighted avg       0.58      0.49      0.48 235435127

Accuracy: 0.4946263604920773

LR_15_0.001

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30469299
           1       0.65      0.86      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.77      0.78      0.76  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7611176899179792
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30469299
           1       0.65      0.86      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.77      0.78      0.76  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7611176899179792

LR_15_0.01

              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379364227336804
              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379364227336804

LR_15_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7039353569180197
              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7039353569180197

LR_15_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202286933929675
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202286933929675

LR_15_0.4

              precision    recall  f1-score   support

           0       0.90      0.63      0.74  26121772
           1       0.57      0.87      0.69  14505950

    accuracy                           0.72  40627722
   macro avg       0.73      0.75      0.71  40627722
weighted avg       0.78      0.72      0.72  40627722

Accuracy: 0.7152346124648583
              precision    recall  f1-score   support

           0       0.90      0.63      0.74  26121772
           1       0.57      0.87      0.69  14505950

    accuracy                           0.72  40627722
   macro avg       0.73      0.75      0.71  40627722
weighted avg       0.78      0.72      0.72  40627722

Accuracy: 0.7152346124648583

LR_15_All

              precision    recall  f1-score   support

           0       0.69      0.32      0.44 144486817
           1       0.42      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.55      0.49 235435127
weighted avg       0.58      0.49      0.48 235435127

Accuracy: 0.49461241609880924
              precision    recall  f1-score   support

           0       0.69      0.32      0.44 144486817
           1       0.42      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.55      0.49 235435127
weighted avg       0.58      0.49      0.48 235435127

Accuracy: 0.49461241609880924

LR_7_0.001

              precision    recall  f1-score   support

           0       0.87      0.67      0.76  30469299
           1       0.63      0.84      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.75      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7383836885270137
              precision    recall  f1-score   support

           0       0.87      0.67      0.76  30469299
           1       0.63      0.84      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.75      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7383836885270137

LR_7_0.01

              precision    recall  f1-score   support

           0       0.87      0.69      0.77  30231476
           1       0.64      0.84      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.748779750217366
              precision    recall  f1-score   support

           0       0.87      0.69      0.77  30231476
           1       0.64      0.84      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.748779750217366

LR_7_0.1

              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7616268733923397
              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7616268733923397

LR_7_0.2

              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7605276006122557
              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7605276006122557

LR_7_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474215758392755
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474215758392755

LR_7_All

              precision    recall  f1-score   support

           0       0.71      0.32      0.44 144486817
           1       0.42      0.79      0.55  90948310

    accuracy                           0.50 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.50      0.49 235435127

Accuracy: 0.5032463571164553
              precision    recall  f1-score   support

           0       0.71      0.32      0.44 144486817
           1       0.42      0.79      0.55  90948310

    accuracy                           0.50 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.50      0.49 235435127

Accuracy: 0.5034408310702081

LR_8_0.001

              precision    recall  f1-score   support

           0       0.86      0.67      0.75  30469299
           1       0.62      0.83      0.71  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.75      0.73  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.73530065264328
              precision    recall  f1-score   support

           0       0.86      0.67      0.75  30469299
           1       0.62      0.83      0.71  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.75      0.73  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.73530065264328

LR_8_0.01

              precision    recall  f1-score   support

           0       0.87      0.69      0.77  30231476
           1       0.64      0.84      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.748819277655453
              precision    recall  f1-score   support

           0       0.87      0.69      0.77  30231476
           1       0.64      0.84      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.748819277655453

LR_8_0.1

              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7610481307173302
              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7610481307173302

LR_8_0.2

              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7600604023818683
              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7600604023818683

LR_8_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474707540826434
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474707540826434

LR_8_All

              precision    recall  f1-score   support

           0       0.71      0.34      0.46 144486817
           1       0.43      0.78      0.55  90948310

    accuracy                           0.51 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.51      0.49 235435127

Accuracy: 0.5086445150557334
              precision    recall  f1-score   support

           0       0.71      0.34      0.46 144486817
           1       0.43      0.78      0.55  90948310

    accuracy                           0.51 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.51      0.49 235435127

Accuracy: 0.5086445150557334

LR_9_0.001

              precision    recall  f1-score   support

           0       0.87      0.67      0.76  30469299
           1       0.63      0.84      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.75      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7390812364416569
              precision    recall  f1-score   support

           0       0.87      0.67      0.76  30469299
           1       0.63      0.84      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.75      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7390812364416569

LR_9_0.01

              precision    recall  f1-score   support

           0       0.87      0.68      0.77  30231476
           1       0.64      0.85      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.7483692483805691
              precision    recall  f1-score   support

           0       0.87      0.68      0.77  30231476
           1       0.64      0.85      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.7483692483805691

LR_9_0.1

              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7608571095693863
              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7608571095693863

LR_9_0.2

              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7598907622213377
              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7598907622213377

LR_9_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474161362037478
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474161362037478

LR_9_All

              precision    recall  f1-score   support

           0       0.71      0.34      0.46 144486817
           1       0.43      0.78      0.55  90948310

    accuracy                           0.51 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.51      0.49 235435127

Accuracy: 0.5085327262995891
              precision    recall  f1-score   support

           0       0.71      0.34      0.46 144486817
           1       0.43      0.78      0.55  90948310

    accuracy                           0.51 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.51      0.49 235435127

Accuracy: 0.5085327262995891

Model Metrics

True Positives

True Negatives

False Positives

False Negatives

Individual Workload Result

zipf_1_15

0.001

0.01

0.1

0.2

0.4

Ignore Obj Size

0.001

0.01

0.1

0.2

0.4

zipf_1_16

0.001

0.01

0.1

0.2

0.4

Ignore Obj Size

0.001

0.01

0.1

0.2

0.4

zipf_1_17

0.001

0.01

0.1

0.2

0.4

Ignore Obj Size

0.001

0.01

0.1

0.2

0.4

zipf_1_18

0.001

0.01

0.1

0.2

0.4

Ignore Obj Size

0.001

0.01

0.1

0.2

0.4

zipf_1_19

0.001

0.01

0.1

0.2

0.4

Ignore Obj Size

0.001

0.01

0.1

0.2

0.4